# Low-Resource Efficient Inference
Arcana Qwen3 2.4B A0.6B
Apache-2.0
This is a Mixture of Experts (MoE) model based on Qwen3, with a total of 2.4 billion parameters, including four expert models with 0.6 billion parameters each, designed to deliver more accurate results with higher efficiency and lower memory usage.
Large Language Model
Transformers Supports Multiple Languages

A
suayptalha
199
21
Falcon H1 34B Instruct GPTQ Int8
Other
Falcon-H1 is a high-performance hybrid architecture language model developed by TII, combining the strengths of Transformers and Mamba architectures, supporting English and multilingual tasks.
Large Language Model
Transformers

F
tiiuae
105
3
Sarvam 1
Sarvam-1 is a 2 billion parameter language model optimized for Indian languages, excelling in 10 Indian languages.
Large Language Model
Transformers Supports Multiple Languages

S
sarvamai
3,386
55
Gemma Portuguese Luana 2b
Apache-2.0
This is a 2B-parameter Portuguese large language model based on the Gemma architecture, specifically optimized for Brazilian Portuguese, supporting instruction-following and text generation tasks.
Large Language Model
Transformers Other

G
rhaymison
115
4
Discolm German 7b V1 AWQ
Apache-2.0
DiscoLM German 7B v1 is a 7B-parameter German language model based on the Mistral architecture, supporting both German and English, and released under the Apache-2.0 license.
Large Language Model
Transformers Supports Multiple Languages

D
TheBloke
81
4
Featured Recommended AI Models